Manifold regularization based on Nystr{ö}m type subsampling

نویسندگان

  • Abhishake Rastogi
  • Sivananthan Sampath
چکیده

In this paper, we study the Nyström type subsampling for large scale kernel methods to reduce the computational complexities of big data. We discuss the multi-penalty regularization scheme based on Nyström type subsampling which is motivated from well-studied manifold regularization schemes. We develop a theoretical analysis of multi-penalty least-square regularization scheme under the general source condition in vector-valued function setting, therefore the results can also be applied to multi-task learning problems. We achieve the optimal minimax convergence rates of multi-penalty regularization using the concept of effective dimension for the appropriate subsampling size. We discuss an aggregation approach based on linear function strategy to combine various Nyström approximants. Finally, we demonstrate the performance of multi-penalty regularization based on Nyström type subsampling on Caltech-101 data set for multi-class image classification and NSL-KDD benchmark data set for intrusion detection problem.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Spectral Partitioning with Inde nite Kernels using the Nystr om Extension

Fowlkes et al recently introduced an approximation to the Normalized Cut NCut grouping algorithm based on random subsampling and the Nystr om extension As presented their method is restricted to the case where W the weighted adjacency matrix is positive de nite Although many common measures of image similarity i e kernels are positive de nite a popular example being Gaussian weighted distance t...

متن کامل

Less is More: Nyström Computational Regularization

We study Nyström type subsampling approaches to large scale kernel methods, and prove learning bounds in the statistical learning setting, where random sampling and high probability estimates are considered. In particular, we prove that these approaches can achieve optimal learning bounds, provided the subsampling level is suitably chosen. These results suggest a simple incremental variant of N...

متن کامل

NYTRO: When Subsampling Meets Early Stopping

Early stopping is a well known approach to reduce the time complexity for performing training and model selection of large scale learning machines. On the other hand, memory/space (rather than time) complexity is the main constraint in many applications, and randomized subsampling techniques have been proposed to tackle this issue. In this paper we ask whether early stopping and subsampling ide...

متن کامل

Habitat structure and body size distributions: cross-ecosystemcomparison for taxa with determinate and indeterminate growth

K. L. Nash ([email protected])(orcid.org/0000-0003-0976-3197) and N. A. J. Graham, ARC Centre of Excellence for Coral Reef Studies, James Cook Univ., Townsville, QLD, 4811, Australia. – C. R. Allen, US Geological Survey Nebraska Cooperative Fish and Wildlife Research Unit, Univ. of Nebraska, Lincoln, NE 68583, USA. – C. Barichievy, Ezemvelo KZN Wildlife, Ithala Game Reserve, Louwsberg 3150, ...

متن کامل

Regularization in Statistics

This paper is a selective review of the regularization methods scattered in statistics literature. We introduce a general conceptual approach to regularization and fit most existing methods into it. We have tried to focus on the importance of regularization when dealing with today’s high-dimensional objects: data and models. A wide range of examples are discussed, including nonparametric regres...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1710.04872  شماره 

صفحات  -

تاریخ انتشار 2017